Using Kullback-Leibler Divergence to Model Opponents in Poker
نویسندگان
چکیده
Opponent modeling is an essential approach for building competitive computer agents in imperfect information games. This paper presents a novel approach to develop opponent modeling techniques. The approach applies neural networks which are separately trained on different dataset to build Kmodel clustering opponent models. KullbackLeibler (KL) divergence is used to exploit a safety mode on opponent modeling. Given a parameter d that controls the max divergence between a model’s centre point and the units belong to it, the approach is proved to provide a lower bound of expected payoff which is above the minimax payoff for correctly clustered players. Even for the players that are incorrectly clustered, the lower bound can also be unlimited approximated with sufficient history data. In our experiments, agent with the novel model shows an improved classification efficiency of opponent modeling comparing with relative researches. And also, the new agent performs better when playing against poker agent HITSZ_CS_13 which participate Annual Computer Poker Competition of 2013.
منابع مشابه
Model Confidence Set Based on Kullback-Leibler Divergence Distance
Consider the problem of estimating true density, h(.) based upon a random sample X1,…, Xn. In general, h(.)is approximated using an appropriate in some sense, see below) model fƟ(x). This article using Vuong's (1989) test along with a collection of k(> 2) non-nested models constructs a set of appropriate models, say model confidence set, for unknown model h(.).Application of such confide...
متن کاملComparison of Kullback-Leibler, Hellinger and LINEX with Quadratic Loss Function in Bayesian Dynamic Linear Models: Forecasting of Real Price of Oil
In this paper we intend to examine the application of Kullback-Leibler, Hellinger and LINEX loss function in Dynamic Linear Model using the real price of oil for 106 years of data from 1913 to 2018 concerning the asymmetric problem in filtering and forecasting. We use DLM form of the basic Hoteling Model under Quadratic loss function, Kullback-Leibler, Hellinger and LINEX trying to address the ...
متن کاملUsing Kullback-Leibler distance for performance evaluation of search designs
This paper considers the search problem, introduced by Srivastava cite{Sr}. This is a model discrimination problem. In the context of search linear models, discrimination ability of search designs has been studied by several researchers. Some criteria have been developed to measure this capability, however, they are restricted in a sense of being able to work for searching only one possibl...
متن کاملInformation Measures via Copula Functions
In applications of differential geometry to problems of parametric inference, the notion of divergence is often used to measure the separation between two parametric densities. Among them, in this paper, we will verify measures such as Kullback-Leibler information, J-divergence, Hellinger distance, -Divergence, … and so on. Properties and results related to distance between probability d...
متن کاملEvaluating the Improvement of Partial Discharge Localization Accuracy Using Frequency Response Assurance Criterion
Partial Discharge (PD) is the most important source of insulation degradation in power transformers. In order to prevent catastrophic failures in transformers, PDs need to be located as soon as possible so that maintenance measures can be taken in time. Due to the structural complexity of windings, locating the PD source inside a transformer winding is not a simple task. In this paper, the effi...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2014